Goto

Collaborating Authors

 deep neural network model


A Statistical Recurrent Model on the Manifold of Symmetric Positive Definite Matrices

Neural Information Processing Systems

In a number of disciplines, the data (e.g., graphs, manifolds) to be analyzed are non-Euclidean in nature. Geometric deep learning corresponds to techniques that generalize deep neural network models to such non-Euclidean spaces. Several recent papers have shown how convolutional neural networks (CNNs) can be extended to learn with graph-based data. In this work, we study the setting where the data (or measurements) are ordered, longitudinal or temporal in nature and live on a Riemannian manifold -- this setting is common in a variety of problems in statistical machine learning, vision and medical imaging. We show how recurrent statistical recurrent network models can be defined in such spaces. We give an efficient algorithm and conduct a rigorous analysis of its statistical properties. We perform extensive numerical experiments demonstrating competitive performance with state of the art methods but with significantly less number of parameters. We also show applications to a statistical analysis task in brain imaging, a regime where deep neural network models have only been utilized in limited ways.


High-contrast "gaudy" images improve the training of deep neural network models of visual cortex

Neural Information Processing Systems

A key challenge in understanding the sensory transformations of the visual system is to obtain a highly predictive model that maps natural images to neural responses. Deep neural networks (DNNs) provide a promising candidate for such a model. However, DNNs require orders of magnitude more training data than neuroscientists can collect because experimental recording time is severely limited. This motivates us to find images that train highly-predictive DNNs with as little training data as possible. We propose high-contrast, binarized versions of natural images---termed gaudy images---to efficiently train DNNs to predict higher-order visual cortical responses. In simulation experiments and analyses of real neural data, we find that training DNNs with gaudy images substantially reduces the number of training images needed to accurately predict responses to natural images. We also find that gaudy images, chosen before training, outperform images chosen during training by active learning algorithms. Thus, gaudy images overemphasize features of natural images that are the most important for efficiently training DNNs. We believe gaudy images will aid in the modeling of visual cortical neurons, potentially opening new scientific questions about visual processing.


Re-optimization of a deep neural network model for electron-carbon scattering using new experimental data

Kowal, Beata E., Graczyk, Krzysztof M., Ankowski, Artur M., Banerjee, Rwik Dharmapal, Bonilla, Jose L., Prasad, Hemant, Sobczyk, Jan T.

arXiv.org Artificial Intelligence

We present an updated deep neural network model for inclusive electron-carbon scattering. Using the bootstrap model [Phys.Rev.C 110 (2024) 2, 025501] as a prior, we incorporate recent experimental data, as well as older measurements in the deep inelastic scattering region, to derive a re-optimized posterior model. We examine the impact of these new inputs on model predictions and associated uncertainties. Finally, we evaluate the resulting cross-section predictions in the kinematic range relevant to the Hyper-Kamiokande and DUNE experiments.


A Statistical Recurrent Model on the Manifold of Symmetric Positive Definite Matrices

Neural Information Processing Systems

In a number of disciplines, the data (e.g., graphs, manifolds) to be analyzed are non-Euclidean in nature. Geometric deep learning corresponds to techniques that generalize deep neural network models to such non-Euclidean spaces. Several recent papers have shown how convolutional neural networks (CNNs) can be extended to learn with graph-based data. In this work, we study the setting where the data (or measurements) are ordered, longitudinal or temporal in nature and live on a Riemannian manifold -- this setting is common in a variety of problems in statistical machine learning, vision and medical imaging. We show how recurrent statistical recurrent network models can be defined in such spaces. We give an efficient algorithm and conduct a rigorous analysis of its statistical properties. We perform extensive numerical experiments demonstrating competitive performance with state of the art methods but with significantly less number of parameters. We also show applications to a statistical analysis task in brain imaging, a regime where deep neural network models have only been utilized in limited ways.


High-contrast "gaudy" images improve the training of deep neural network models of visual cortex

Neural Information Processing Systems

A key challenge in understanding the sensory transformations of the visual system is to obtain a highly predictive model that maps natural images to neural responses. Deep neural networks (DNNs) provide a promising candidate for such a model. However, DNNs require orders of magnitude more training data than neuroscientists can collect because experimental recording time is severely limited. This motivates us to find images that train highly-predictive DNNs with as little training data as possible. We propose high-contrast, binarized versions of natural images---termed gaudy images---to efficiently train DNNs to predict higher-order visual cortical responses.


Review for NeurIPS paper: High-contrast "gaudy" images improve the training of deep neural network models of visual cortex

Neural Information Processing Systems

Weaknesses: The main weakness of the work to me was the context in which the authors proposed and evaluated the use of gaudy images. Modeling neural activity using DNNs is of great importance in computational neuroscience. However, the authors motivate this specific method because they claim there is not enough neural data to train complex models on normal natural images, either straight from image to neural activity or from pre-trained DNN activities to neural activity. I find both of these claims misleading: there is quite a bit of work successfully training end-to-end deep networks from images to neural predictions even with limited experimental data. I especially object to the author's explanation that linear mappings are used between pre-trained DNNs and neural activity due to lack of data.


Review for NeurIPS paper: High-contrast "gaudy" images improve the training of deep neural network models of visual cortex

Neural Information Processing Systems

This paper had borderline scores. Overall, I think this paper presents a nice core finding that was sufficiently well validated in the context of simulations. The simulated results are reasonably compelling and relatively thorough analyses were presented. In the discussion with reviewers, there was a reasonable consensus that the author response overemphasized the extent to which the reviewers were hung up on the lack of experimental data. While R3 felt most strongly that this specific paper was not strong enough without further empirical validation, this was clarified to not be a bias against simulation papers generally, but rather the reviewer's opinion that they were not convinced this specific paper's results were strong enough without real data.


High-contrast "gaudy" images improve the training of deep neural network models of visual cortex

Neural Information Processing Systems

A key challenge in understanding the sensory transformations of the visual system is to obtain a highly predictive model that maps natural images to neural responses. Deep neural networks (DNNs) provide a promising candidate for such a model. However, DNNs require orders of magnitude more training data than neuroscientists can collect because experimental recording time is severely limited. This motivates us to find images that train highly-predictive DNNs with as little training data as possible. We propose high-contrast, binarized versions of natural images---termed gaudy images---to efficiently train DNNs to predict higher-order visual cortical responses.


Developing an Explainable Artificial Intelligent (XAI) Model for Predicting Pile Driving Vibrations in Bangkok's Subsoil

Youwai, Sompote, Pamungmoon, Anuwat

arXiv.org Artificial Intelligence

This study presents an explainable artificial intelligent (XAI) model for predicting pile driving vibrations in Bangkok's soft clay subsoil. A deep neural network was developed using a dataset of 1,018 real-world pile driving measurements, encompassing variations in pile dimensions, hammer characteristics, sensor locations, and vibration measurement axes. The model achieved a mean absolute error (MAE) of 0.276, outperforming traditional empirical methods and other machine learning approaches such as XGBoost and CatBoost. SHapley Additive exPlanations (SHAP) analysis was employed to interpret the model's predictions, revealing complex relationships between input features and peak particle velocity (PPV). Distance from the pile driving location emerged as the most influential factor, followed by hammer weight and pile size. Non-linear relationships and threshold effects were observed, providing new insights into vibration propagation in soft clay. A web-based application was developed to facilitate adoption by practicing engineers, bridging the gap between advanced machine learning techniques and practical engineering applications. This research contributes to the field of geotechnical engineering by offering a more accurate and nuanced approach to predicting pile driving vibrations, with implications for optimizing construction practices and mitigating environmental impacts in urban areas. The model and its source code are publicly available, promoting transparency and reproducibility in geotechnical research.


MSMix:An Interpolation-Based Text Data Augmentation Method Manifold Swap Mixup

Ye, Mao, Wang, Haitao, Chen, Zheqian

arXiv.org Artificial Intelligence

To solve the problem of poor performance of deep neural network models due to insufficient data, a simple yet effective interpolation-based data augmentation method is proposed: MSMix (Manifold Swap Mixup). This method feeds two different samples to the same deep neural network model, and then randomly select a specific layer and partially replace hidden features at that layer of one of the samples by the counterpart of the other. The mixed hidden features are fed to the model and go through the rest of the network. Two different selection strategies are also proposed to obtain richer hidden representation. Experiments are conducted on three Chinese intention recognition datasets, and the results show that the MSMix method achieves better results than other methods in both full-sample and small-sample configurations.